Guest editorial: Assessing the cost of data quality in a ?low bid? world

Author(s):  
Thomas L. Francoeur
Author(s):  
Hatice Uenal ◽  
David Hampel

Registries are indispensable in medical studies and provide the basis for reliable study results for research questions. Depending on the purpose of use, a high quality of data is a prerequisite. However, with increasing registry quality, costs also increase accordingly. Considering these time and cost factors, this work is an attempt to estimate the cost advantages of applying statistical tools to existing registry data, including quality evaluation. Results for quality analysis showed that there are unquestionable savings of millions in study costs by reducing the time horizon and saving on average € 523,126 for every reduced year. Replacing additionally the over 25 % missing data in some variables, data quality was immensely improved. To conclude, our findings showed dearly the importance of data quality and statistical input in avoiding biased conclusions due to incomplete data.


2010 ◽  
Author(s):  
Peter Napier ◽  
Ian M Threadgold ◽  
Per Gunnar Aas ◽  
David J Harrison
Keyword(s):  

2020 ◽  
Vol 13 (11) ◽  
pp. 6237-6254
Author(s):  
Leonardo Alcayaga

Abstract. Wind lidars present advantages over meteorological masts, including simultaneous multipoint observations, flexibility in measuring geometry, and reduced installation cost. But wind lidars come with the “`cost” of increased complexity in terms of data quality and analysis. Carrier-to-noise ratio (CNR) has been the metric most commonly used to recover reliable observations from lidar measurements but with severely reduced data recovery. In this work we apply a clustering technique to identify unreliable measurements from pulsed lidars scanning a horizontal plane, taking advantage of all data available from the lidars – not only CNR but also line-of-sight wind speed (VLOS), spatial position, and VLOS smoothness. The performance of this data filtering technique is evaluated in terms of data recovery and data quality against both a median-like filter and a pure CNR-threshold filter. The results show that the clustering filter is capable of recovering more reliable data in noisy regions of the scans, increasing the data recovery up to 38 % and reducing by at least two-thirds the acceptance of unreliable measurements relative to the commonly used CNR threshold. Along with this, the need for user intervention in the setup of data filtering is reduced considerably, which is a step towards a more automated and robust filter.


2020 ◽  
Author(s):  
Leonardo Alcayaga

Abstract. Wind lidars present advantages over meteorological masts, including simultaneous multi-point observations, flexibility in measuring geometry, and reduced installation cost; but wind lidars come with the cost of increased complexity in terms of data quality and analysis. Carrier-to-noise ratio (CNR) has been the metric most commonly-used to recover reliable observations from lidar measurements, but with severely reduced data recovery. In this work we apply a clustering technique to identify unreliable measurements from pulsed lidars scanning a horizontal plane, taking advantage of all data available from the lidars–not only CNR, but also line-of-sight wind speed (VLOS), spatial position, and VLOS smoothness. The performance of this data filtering technique is evaluated in terms of data recovery and data quality, against both a median-like filter and a pure CNR-threshold filter. The results show that the clustering filter is capable of recovering more reliable data in noisy regions of the scans, increasing the data recovery up to 38 % and reducing by at least two thirds the acceptance of unreliable measurements, relative to the commonly used CNR-threshold. Along with this, the need for user intervention in the setup of data filtering is reduced considerably, which is a step towards a more automated and robust filter.


2019 ◽  
Vol 181 ◽  
pp. 104954 ◽  
Author(s):  
Carlos Sáez ◽  
Siaw-Teng Liaw ◽  
Eizen Kimura ◽  
Pascal Coorevits ◽  
Juan M Garcia-Gomez

2017 ◽  
Vol 50 (4) ◽  
pp. 1005-1036 ◽  
Author(s):  
Charles Breton ◽  
Fred Cutler ◽  
Sarah Lachance ◽  
Alex Mierke-Zatwarnicki

AbstractElection studies must optimize on sample size, cost and data quality. The 2015 Canadian Election Study was the first CES to employ a full mixed-mode design, aiming to take advantage of the opportunities of each mode while preserving enough commonality to compare them. This paper examines the phone interviews conducted by ISR-York and the online questionnaires from panellists purchased from a sample provider. We compare data quality and representativeness. We conduct a comprehensive comparison of the distributions of responses across modes and a comparative analysis of inferences about voting. We find that the cost/power advantages of the online mode will likely make it the mode of choice for subsequent election studies.


2021 ◽  
Author(s):  
Herbert Mauch ◽  
Jasmin Kaur ◽  
Colin Irwin ◽  
Josie Wyss

Abstract Background Registries are powerful clinical investigational tools. More challenging, however, is an international registry conducted by industry. That requires considerable planning, clear objectives and endpoints, resources and appropriate measurement tools. Methods This paper aims to summarize our learning from ten years of running a medical device registry monitoring patient-reported benefits from hearing implants. Results We enlisted 113 participating clinics globally, resulting in a total enrolment of more than 1500 hearing-implant users. We identify the stages in developing a registry specific to a sensory handicap such as hearing loss, its challenges and successes in design and implementation, and recommendations for future registries. Conclusions Data collection infrastructure needs to be maintained up to date throughout the defined registry lifetime and provide adequate oversight of data quality and completeness. Compliance at registry sites is important for data quality and needs to be weighed against the cost of site monitoring. To motivate sites to provide accurate and timely data entry we facilitated easy access to their own data which helped to support their clinical routine. Trial registration: ClinicalTrials.gov NCT02004353


2020 ◽  
Author(s):  
Kris Villez ◽  
Peter A Vanrolleghem ◽  
lluis corominas

The advent of affordable computing, low-cost sensor hardware, and high-speed and reliable communications have spurred ubiquitous installation of sensors in complex engineered systems. However, ensuring reliable data quality remains a challenge. Exploitation of redundancy among sensor signals can help improving the precision of measured variables, detecting the presence of gross errors, and identifying faulty sensors. The cost of sensor ownership, maintenance efforts in particular, can still be cost-prohibitive however. Maximizing the ability to assess and control data quality while minimizing the cost of ownership thus requires a careful sensor placement. To solve this challenge, we develop a generally applicable method to solve the multi-objective sensor placement problem in systems governed by linear and bilinear balance equations. Importantly, the method computes all Pareto-optimal sensor layouts with conventional computational resources and requires no information about the expected sensor quality.


GeoArabia ◽  
1996 ◽  
Vol 1 (4) ◽  
pp. 511-528
Author(s):  
Richard Hastings-James ◽  
Kamal M. Al-Yahya

ABSTRACT Between 1991 and 1996, Saudi Aramco has acquired more than 8,500 square kilometers of 3-D seismic data in Saudi Arabia. During this time, a universal approach to 3-D acquisition has been developed. The resulting acquisition schemes use a dense source point grid with a low sweep effort per source point, and a high number of recorded channels distributed over a large surface aperture. This sampling strategy results in high fold data. Cost-effectiveness is achieved by ensuring that the source and receiver effort are balanced. Comparisons have shown that increases in surface aperture and fold, cross-line fold in particular, improve the data quality significantly at a marginal increase in cost. The cost per unit of data is made significantly lower even if the cost per unit of time may increase.


Sign in / Sign up

Export Citation Format

Share Document